Bayesian dropout

نویسندگان

چکیده

Dropout has recently emerged as a powerful and simple method for training neural networks preventing co-adaptation by stochastically omitting neurons. is currently not grounded in explicit modelling assumptions which so far precluded its adoption Bayesian modelling. Using entropic reasoning we show that dropout can be interpreted optimal inference under constraints. We demonstrate this on an analytically tractable regression model providing interpretation of mechanism regularizing well connection to other techniques. also discuss two general approximate techniques applying models, one based analytical approximation the stochastic variational These are then applied Baysian logistic problem shown improve performance become more misspecified. Our framework roots theoretically justified practical tool statistical allowing Bayesians tap into benefits training.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Bayesian encourages dropout

Dropout is one of the key techniques to prevent the learning from overfitting. It is explained that dropout works as a kind of modified L2 regularization. Here, we shed light on the dropout from Bayesian standpoint. Bayesian interpretation enables us to optimize the dropout rate, which is beneficial for learning of weight parameters and prediction after learning. The experiment result also enco...

متن کامل

Dropout as a Bayesian Approximation: Appendix

We show that a neural network with arbitrary depth and non-linearities, with dropout applied before every weight layer, is mathematically equivalent to an approximation to a well known Bayesian model. This interpretation might offer an explanation to some of dropout’s key properties, such as its robustness to overfitting. Our interpretation allows us to reason about uncertainty in deep learning...

متن کامل

High Dimensional Bayesian Optimization using Dropout

Scaling Bayesian optimization to high dimensions is challenging task as the global optimization of high-dimensional acquisition function can be expensive and often infeasible. Existing methods depend either on limited “active” variables or the additive form of the objective function. We propose a new method for high-dimensional Bayesian optimization, that uses a dropout strategy to optimize onl...

متن کامل

Dropout Inference in Bayesian Neural Networks with Alpha-divergences

To obtain uncertainty estimates with real-world Bayesian deep learning models, practical inference approximations are needed. Dropout variational inference (VI) for example has been used for machine vision and medical applications, but VI can severely underestimates model uncertainty. Alpha-divergences are alternative divergences to VI’s KL objective, which are able to avoid VI’s uncertainty un...

متن کامل

A PAC-Bayesian Tutorial with A Dropout Bound

This tutorial gives a concise overview of existing PAC-Bayesian theory focusing on three generalization bounds. The first is an Occam bound which handles rules with finite precision parameters and which states that generalization loss is near training loss when the number of bits needed to write the rule is small compared to the sample size. The second is a PAC-Bayesian bound providing a genera...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Procedia Computer Science

سال: 2022

ISSN: ['1877-0509']

DOI: https://doi.org/10.1016/j.procs.2022.03.105